Goto

Collaborating Authors

 optimal black-box reduction


Optimal Black-Box Reductions Between Optimization Objectives

Neural Information Processing Systems

The diverse world of machine learning applications has given rise to a plethora of algorithms and optimization methods, finely tuned to the specific regression or classification task at hand. We reduce the complexity of algorithm design for machine learning by reductions: we develop reductions that take a method developed for one setting and apply it to the entire spectrum of smoothness and strong-convexity in applications. Furthermore, unlike existing results, our new reductions are OPTIMAL and more PRACTICAL. We show how these new reductions give rise to new and faster running times on training linear classifiers for various families of loss functions, and conclude with experiments showing their successes also in practice.


Reviews: Optimal Black-Box Reductions Between Optimization Objectives

Neural Information Processing Systems

Overall, the paper is convincing and well written. The problem is very relevant for machine learning applications and the proof arguments are quite convincing. I could not find major flaw in the paper. I have several concerns to raise regarding this work. Smoothing has a long history in optimization and recent work on accelerated methods through strong convexity are not properly presented.

  Genre: Summary/Review (0.36)
  Industry: Transportation > Air (0.40)

Optimal Black-Box Reductions Between Optimization Objectives

Allen-Zhu, Zeyuan, Hazan, Elad

Neural Information Processing Systems

The diverse world of machine learning applications has given rise to a plethora of algorithms and optimization methods, finely tuned to the specific regression or classification task at hand. We reduce the complexity of algorithm design for machine learning by reductions: we develop reductions that take a method developed for one setting and apply it to the entire spectrum of smoothness and strong-convexity in applications. Furthermore, unlike existing results, our new reductions are OPTIMAL and more PRACTICAL. We show how these new reductions give rise to new and faster running times on training linear classifiers for various families of loss functions, and conclude with experiments showing their successes also in practice. Papers published at the Neural Information Processing Systems Conference.